Placing Feedforward Neural Networks Among Several Circuit Complexity Classes

نویسندگان

  • Valeriu Beiu
  • Jan A. Peperstraete
  • Joos Vandewalle
  • Rudy Lauwereins
چکیده

This paper examines the circuit complexity of feedforward neural networks having sigmoid activation function. The starting point is the complexity class NN defined in [18]. First two additional complexity classes NN∆ k and NN∆,ε k having less restrictive conditions (than NN) concerning fan-in and accuracy are defined. We then prove several relations among these three classes and well established circuit complexity classes. All the proofs are constructive as being built around binary adders. By relaxing the fan-in condition from logarithmic [18] to (almost) polynomial, the results show that substantially larger classes of sigmoid activation feedforward neural networks can be implemented in polynomial size Boolean circuits. This is done at the expense of a logarithmic factor increase in the number of layers, but having constant fan-in (of 2). The main conclusion is that there are interesting fan-in dependent depth-size tradeoffs when trying to digitally implement sigmoid activation neural networks.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Classes of feedforward neural networks and their circuit complexity

-Th& paper aims to p&ce neural networks in the conte.\t ol'booh'an citz'ldt complexit.l: 1,1~, de/itte aplm~priate classes qlfeedybrward neural networks with specified fan-in, accm'ac)' olcomputation and depth and ttsing techniques" o./commzmication comph:¥ity proceed to show t/tat the classes.fit into a well-studied hieralz'h)' q/boolean circuits. Results cover both classes of sigmoid activati...

متن کامل

Circuit Complexity and Feedforward Neural Networks

Circuit complexity, a subfield of computational complexity theory, can be used to analyze how the resource usage of neural networks scales with problem size. The computational complexity of discrete feedforward neural networks is surveyed, with a comparison of classical circuits to circuits constructed from gates that compute weighted majority functions.

متن کامل

On Properties of Networks of Neuron-Like Elements

The complexity and computational capacity of multi-layered, feedforward neural networks is examined. Neural networks for special purpose (structured) functions are examined from the perspective of circuit complexity. Known results in complexity theory are applied to the special instance of neural network circuits, and in particular, classes of functions that can be implemented in shallow circui...

متن کامل

Complexity of Predictive Neural Networks

We announce some general bounds on the interactions of neural and information complexities of feedforward neural networks using general classes of activation functions. We show that, up to constant factors, neural and information complexities combine in a well-defined way in the determination of the complexity of a network.

متن کامل

Hardness Results for General Two-Layer Neural Networks

We deal with the problem of learning a general class of 2-layer neural networks in polynomial time. The considered neural networks consist of k linear threshold units on the hidden layer and an arbitrary binary output unit. We show NP-completeness of the consistency problem for classes that use an arbitrary set of binary output units containing a function which depends on all input dimensions. ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007